A Unified SMT Framework Combining MIRA and MERT
نویسندگان
چکیده
Translation sub-model is one of the most important components in statistical machine translation, but the conventional approach suffers from two major problems. Firstly, translation sub-model is not optimized with respect to any of automatic evaluation metrics of SMT (such as BLEU). The second problem is over-fitting to training data. This paper presents a new unified framework, by adding a scalable translation sub-model into the conventional framework. The sub-model is optimized with the same criterion as the translation output is evaluated (BLEU), and trained using margin infused relaxed algorithm (MIRA) to handle over-fitting. Under our new framework, MIRA and minimum error rate training (MERT) are unified into an interactive training process. Our approach has not only shown to improve performance over a state-of-the-art baseline, but also generalize well in-domain training data to out-of-domain test data.
منابع مشابه
Margin Infused Relaxed Algorithm for Moses
We describe an open-source implementation of the Margin Infused Relaxed Algorithm (MIRA) for statistical machine translation (SMT). The implementation is part of the Moses toolkit and can be used as an alternative to standard minimum error rate training (MERT). A description of the implementation and its usage on core feature sets as well as large, sparse feature sets is given and we report exp...
متن کاملBatch Tuning Strategies for Statistical Machine Translation
There has been a proliferation of recent work on SMT tuning algorithms capable of handling larger feature sets than the traditional MERT approach. We analyze a number of these algorithms in terms of their sentencelevel loss functions, which motivates several new approaches, including a Structured SVM. We perform empirical comparisons of eight different tuning strategies, including MERT, in a va...
متن کاملTuning Methods in Statistical Machine Translation
In a Statistical Machine Translation system many models, called features, complement each other in producing natural language translations. In how far we should rely on a certain feature is governed by parameters, or weights. Learning these weights is the sub field of SMT, called parameter tuning, that is addressed in this thesis. Three existing methods for learning such parameters are compared...
متن کاملSearch-Aware Tuning for Hierarchical Phrase-based Decoding
Parameter tuning is a key problem for statistical machine translation (SMT). Most popular parameter tuning algorithms for SMT are agnostic of decoding, resulting in parameters vulnerable to search errors in decoding. The recent research of “search-aware tuning” (Liu and Huang, 2014) addresses this problem by considering the partial derivations in every decoding step so that the promising ones a...
متن کاملTuning as Ranking
We offer a simple, effective, and scalable method for statistical machine translation parameter tuning based on the pairwise approach to ranking (Herbrich et al., 1999). Unlike the popular MERT algorithm (Och, 2003), our pairwise ranking optimization (PRO) method is not limited to a handful of parameters and can easily handle systems with thousands of features. Moreover, unlike recent approache...
متن کامل